Web Survey Bibliography
As the social and economic problems of the United States become more complex -- AIDS, sexual abuse, violence, homelessness, corruption-- social scientists are called upon more frequently to identify and understand behaviors that are private and sensitive. Survey researchers have attempted to obtain this information by asking people to report their own behaviors in interviews, and have concluded, not surprisingly, that the more private the forum for reporting, the more accurate the report (Bradburn and Sudman, 1979; Bradburn, 1983). Moreover, in demonstrated experiments, the self-administered interview has been determined to provide the best environment for the reporting of many types of sensitive behaviors (Hay, 1990; London and Williams, 1990; Aquilino and LoSciuto, 1990; Schwarz, et al, 1991; Turner, et al, 1992). Limitations of the self-administered interview, however, are obvious. The respondent must have adequate reading skills, and must be able to understand and follow the questionnaire format, a novel experience for many (Lessler and Holt, 1987). Consequently, the burden on the researcher is to design a questionnaire void of complexity yet accomplishing the research objectives. Solutions to this dilemma have been evolving. The successful use of a Walkman-type device in the Youth Risk Behavior Survey conducted by the National Center of Health Statistics (NCHS) (Camburn, et al, 1991) eliminated the need for adequate reading skills by the respondent. But this method could not use skip instructions to access questions, thereby restricting the complexity of the instrument. Incorporating audio with a computerized selfadministration of a questionnaire (audio computerassisted self interviewing, or Audio-CASI) has offered an improved solution. With Audio-CASI the respondent listens to a voice-digitized recording of the questions and answer choices over earphones and keys the answers into a microcomputer. While providing privacy and eliminating the need for respondent literacy, this technology also allows the researcher to design complex questionnaires and provides standardized questionnaire administration.
Homepage - conference (full text)
Web survey bibliography - Conference proceedings (83)
- The Influence of Human Cues on Drop-out and Answer Behaviour within Web Surveys; 2008; Oesterlau, T., Geißler, H.
- Sampling for web surveys; 2007; Rivers, D.
- Reconstructing childhood health histories using internet panels; 2007; Smith, J. P.
- Pilot study to recruite a sample for an online panel: Effects of contact mode, incentives and information...; 2007; Scherpenzeel, A.
- Modes, trends, and content: A comparison of the 2003 HRS internet survey with HRS 2002 and 2004 Core...; 2007; Weir, D.
- Lessons learned: Converting a telephone survey panel to an internet panel; 2007; Roe, D. J., Stockdale, J., Farrelly, M., Heinrich, T.
- Developments in electronic survey design for establishment surveys; 2007; O'Neill, G.
- Response time measurement in the lab and on the Web: A comparison; 2007; Galesic, M., Reips, U.-D., Kaczmirek, L., Czienskowski, U., Liske, N., von Oertzen, T.
- Questionnaire and Survey Design for Online Research; 2007; Wydra, D., Fisher, L., Strunk, K.
- Using Technology to Improve Data Quality; 2007; Cunningham, J., Pearson, C., Parkhurst, D.
- Mobile Phone Data Collection; 2007; Lavine, S.
- Effects of Offering Web Questionnaires as an Option in Enterprise Surveys: the Swedish Experience; 2007; Erikson, J., Furubjelke, E.
- Using the Web for Surveys of Medical Providers; 2007; Narayanan, V., Giambo, P., Fry, S., Crafts, J.
- Boosting Response Rates: Are There Optimal Times to Email Respondents?; 2007; Bennett-Harper, S., O'Brien, J., Levin, K., Davis, B., Shipp, S., Campbell, S., Sienkiewicz, R.
- Beyond Demographics: Are ‘Webographic’ Questions Useful for Reducing the Selection Bias...; 2007; Schonlau, M., van Soest, A., Kapteyn, A.
- Maximising respondent engagement: The use of rich media; 2007; Reid, J., Morden, M., Perez Echeverria, A.
- Truth in measurement: Comparing Web Based interviewing Techniques; 2007; Couper, M. P., Terhanian, G., Bremer, J., Thomas, R. K.
- The power of the visible: Visual design for Web surveys; 2006; Couper, M. P.
- Attitudinal differences. Comparing people who belong to multiple versus single panels; 2006; Casdas, D., Fine, B., Menictas, C.
- Weighting an Internet Panel Survey on Drug Use and Abuse; 2006; Gordek, H., Williams, Ri. L., Dai, L.
- The professional respondent problem in online panel surveys today; 2005; Fulgoni, G.
- Satisficing behavior in online panelists; 2005; Downes-Le Guin, T.
- Panel proliferation and quality concerns; 2005; Faasse, J.
- Electronic Voting Machines – A comparison applying the principles of computer-human interaction...; 2003; Callegaro, M., Peytcheva, E.
- More than a thousand words? Visual cues and visual knowledge; 2002; Prior, M.
- Questionnaire Pretesting Methods: Do Different Techniques and Different Organizations Produce Similar...; 2001; Rothgeb, J. M., Willis, G. B., Forsyth, B. H.
- Practical methods for sampling rare and mobile populations; 2001; Kalton, G.
- Building an alternative response process model for business surveys; 2001; Willimack, D. K., Nichols, E. M.
- Human centered measures of success in web site design; 1998; Kirakowski, J., Claridge, N., Whitehand, R.
- The Prodigy Experiment in Using e-Mail for Tracking Public Opinion; 1995; Werner, J., Maisel, R., Robinson, K.
- Respondent preferences toward audio-CASI and how that affects data quality; 1995; Kinsey, S. H., Thornberry, J. S., Carson, C. P., Duffer, A. P.
- Best pracices in disk-by-mail surveys; 1992; Witt, K. J., Bernstein, S.
- Customer satisfaction research using disks-by-mail; 1989; Zabdan, P., Frost, L.